Multi-Conditional Learning for Joint Probability Models with Latent Variables
نویسندگان
چکیده
We introduce Multi-Conditional Learning, a framework for optimizing graphical models based not on joint likelihood, or on conditional likelihood, but based on a product of several marginal conditional likelihoods each relying on common sets of parameters from an underlying joint model and predicting different subsets of variables conditioned on other subsets. When applied to undirected models with latent variables, such as the Harmonium, this approach can result in powerful, structured latent variable representations that combine some of the advantages of conditional random fields with the unsupervised clustering ability of popular topic models, such as latent Dirichlet allocation and its successors. We present new algorithms for parameter estimation using expected gradient based optimization and develop fast approximate inference algorithms inspired by the contrastive divergence approach. Our initial experimental results show improved cluster quality on synthetic data, promising results on a vowel recognition problem and significant improvement inferring hidden document categories from multiple attributes of documents.
منابع مشابه
Predictive Random Fields: Latent Variable Models Fit by Multiway Conditional Probability with Applications to Document Analysis
We introduce predictive random fields, a framework for learning undirected graphical models based not on joint, generative likelihood, or on conditional likelihood, but based on a product of several conditional likelihoods each relying on common sets of parameters and predicting different subsets of variables conditioned on other subsets. When applied to models with latent variables, such as th...
متن کاملPosterior regularization for Joint Modelling of Multiple Structured Prediction Tasks with Soft Constraints
We propose a multi-task learning objective for training joint structured prediction models when no jointly annotated data is available. We use conditional random fields as the joint predictive model and train their parameters by optimizing the marginal likelihood of all available annotations, with additional posterior constraints on the distributions of the latent variables imposed to enforce a...
متن کاملMulti-Conditional Learning: Generative/Discriminative Training for Clustering and Classification
This paper presents multi-conditional learning (MCL), a training criterion based on a product of multiple conditional likelihoods. When combining the traditional conditional probability of “label given input” with a generative probability of “input given label” the later acts as a surprisingly effective regularizer. When applied to models with latent variables, MCL combines the structure-discov...
متن کاملCumulative distribution networks: Inference, estimation and applications of graphical models for cumulative distribution functions
Cumulative distribution networks: Inference, estimation and applications of graphical models for cumulative distribution functions Jim C. Huang Doctor of Philosophy Graduate Department of Electrical and Computer Engineering University of Toronto 2009 This thesis presents a class of graphical models for directly representing the joint cumulative distribution function (CDF) of many random variabl...
متن کاملCS 6782: Fall 2010 Probabilistic Graphical Models
In a probabilistic graphical model, each node represents a random variable, and the links express probabilistic relationships between these variables. The structure that graphical models exploit is the independence properties that exist in many real-world phenomena. The graph then captures the way in which the joint distribution over all of the random variables can be decomposed into a product ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2005